# Wikipedia corpus

Multilingual Albert Base Cased 128k
Apache-2.0
A multilingual ALBERT model pretrained with masked language modeling (MLM) objective, supporting 60+ languages, featuring a lightweight architecture with parameter sharing
Large Language Model Transformers Supports Multiple Languages
M
cservan
277
2
Multilingual Albert Base Cased 32k
Apache-2.0
Multilingual ALBERT model pretrained with masked language modeling objective, supporting 50+ languages, case-sensitive
Large Language Model Transformers Supports Multiple Languages
M
cservan
243
2
Tinybert General 4L 312D De
This is a TinyBERT model optimized for German, created by distilling the BERT base cased model, suitable for natural language processing tasks.
Large Language Model Transformers German
T
dvm1983
269
3
Gpt2 Small Indonesian 522M
MIT
This is a GPT2-small model pretrained on Indonesian Wikipedia data, specializing in Indonesian text generation tasks.
Large Language Model Other
G
cahya
1,900
9
Slovakbert
MIT
A pretrained model based on Slovak language, using masked language modeling (MLM) objective, case-sensitive.
Large Language Model Transformers Other
S
gerulata
5,009
23
Gpt2 Small Arabic
A GPT2 model trained on Arabic Wikipedia dataset based on gpt2-small architecture, suitable for text generation tasks.
Large Language Model Arabic
G
akhooli
1,066
15
Electra Small Japanese Generator
An ELECTRA model pre-trained on Japanese Wikipedia, suitable for Japanese text processing tasks.
Large Language Model Transformers Japanese
E
izumi-lab
16
0
Bert Base Japanese Upos
BERT model pre-trained on Japanese Wikipedia text for POS tagging and dependency parsing
Sequence Labeling Transformers Supports Multiple Languages
B
KoichiYasuoka
40
2
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase